85 research outputs found
Improving the performance of object detection by preserving label distribution
Object detection is a task that performs position identification and label
classification of objects in images or videos. The information obtained through
this process plays an essential role in various tasks in the field of computer
vision. In object detection, the data utilized for training and validation
typically originate from public datasets that are well-balanced in terms of the
number of objects ascribed to each class in an image. However, in real-world
scenarios, handling datasets with much greater class imbalance, i.e., very
different numbers of objects for each class , is much more common, and this
imbalance may reduce the performance of object detection when predicting unseen
test images. In our study, thus, we propose a method that evenly distributes
the classes in an image for training and validation, solving the class
imbalance problem in object detection. Our proposed method aims to maintain a
uniform class distribution through multi-label stratification. We tested our
proposed method not only on public datasets that typically exhibit balanced
class distribution but also on custom datasets that may have imbalanced class
distribution. We found that our proposed method was more effective on datasets
containing severe imbalance and less data. Our findings indicate that the
proposed method can be effectively used on datasets with substantially
imbalanced class distribution.Comment: Code is available at
https://github.com/leeheewon-01/YOLOstratifiedKFold/tree/mai
Globally Convergent Ordered Subsets Algorithms: Application to Tomography
We present new algorithms for penalized-likelihood image reconstruction: modified BSREM (block sequential regularized expectation maximization) and relaxed OS-SPS (ordered subsets separable paraboloidal surrogates). Both of them are globally convergent to the unique solution, easily incorporate convex penalty functions, and are parallelizable-updating all voxels (or pixels) simultaneously. They belong to a class of relaxed ordered subsets algorithms. We modify the scaling function of the existing BSREM (De Pierro and Yamagishi, 2001) so that we can prove global convergence without previously imposed assumptions. We also introduce a diminishing relaxation parameter into the existing OS-SPS (Erdogan and Fessler, 1999) to achieve global convergence. We also modify the penalized-likelihood function to enable the algorithms to cover a zero-background-event case. Simulation results show that the algorithms are both globally convergent and fast.Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/86018/1/Fessler168.pd
Statistical Emission Image Reconstruction for Randoms-Precorrected PET Scans Using Negative Sinogram Values
Many conventional PET emission scans are corrected for accidental coincidence (AC) events, or randoms, by real-time subtraction of delayed-window coincidences, leaving only the randoms-precorrected data available for image reconstruction. The real-time precorrection compensates in mean for AC events but destroys Poisson statistics. Since the exact log-likelihood for randoms-precorrected data is inconvenient to maximize, practical approximations are desirable for statistical image reconstruction. Conventional approximations involve setting negative sinogram values to zero, which can induce positive systematic biases, particularly for scans with low counts per ray. We propose new likelihood approximations that allow negative sinogram values without requiring zero-thresholding. We also develop monotonic algorithms for the new models by using "optimization transfer" principles. Simulation results show that our new model, SP-, is free of systematic bias yet keeps low variance. Despite its simpler implementation, the new model performs comparably to the saddle-point (SD) model which has previously shown the best performance (as to systematic bias and variance) in randoms-precorrected PET emission reconstruction.Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/85893/1/Fessler185.pd
Globally Convergent Image Reconstruction for Emission Tomography Using Relaxed Ordered Subsets Algorithms
We present two types of globally convergent relaxed ordered subsets (OS) algorithms for penalized-likelihood image reconstruction in emission tomography: modified block sequential regularized expectation-maximization (BSREM) and relaxed OS separable paraboloidal surrogates (OS-SPS). The global convergence proof of the existing BSREM (De Pierro and Yamagishi, 2001) required a few a posteriori assumptions. By modifying the scaling functions of BSREM, we are able to prove the convergence of the modified BSREM under realistic assumptions. Our modification also makes stepsize selection more convenient. In addition, we introduce relaxation into the OS-SPS algorithm (Erdogan and Fessler, 1999) that otherwise would converge to a limit cycle. We prove the global convergence of diagonally scaled incremental gradient methods of which the relaxed OS-SPS is a special case; main results of the proofs are from (Nedic and Bertsekas, 2001) and (Correa and Lemarechal, 1993). Simulation results showed that both new algorithms achieve global convergence yet retain the fast initial convergence speed of conventional unrelaxed ordered subsets algorithms.Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/86017/1/Fessler67.pd
Recommended from our members
Quantitative Statistical Methods for Image Quality Assessment
Quantitative measures of image quality and reliability are critical for both qualitative interpretation and quantitative analysis of medical images. While, in theory, it is possible to analyze reconstructed images by means of Monte Carlo simulations using a large number of noise realizations, the associated computational burden makes this approach impractical. Additionally, this approach is less meaningful in clinical scenarios, where multiple noise realizations are generally unavailable. The practical alternative is to compute closed-form analytical expressions for image quality measures. The objective of this paper is to review statistical analysis techniques that enable us to compute two key metrics: resolution (determined from the local impulse response) and covariance. The underlying methods include fixed-point approaches, which compute these metrics at a fixed point (the unique and stable solution) independent of the iterative algorithm employed, and iteration-based approaches, which yield results that are dependent on the algorithm, initialization, and number of iterations. We also explore extensions of some of these methods to a range of special contexts, including dynamic and motion-compensated image reconstruction. While most of the discussed techniques were developed for emission tomography, the general methods are extensible to other imaging modalities as well. In addition to enabling image characterization, these analysis techniques allow us to control and enhance imaging system performance. We review practical applications where performance improvement is achieved by applying these ideas to the contexts of both hardware (optimizing scanner design) and image reconstruction (designing regularization functions that produce uniform resolution or maximize task-specific figures of merit)
Emission Image Reconstruction for Randoms-Precorrected PET Allowing Negative Sinogram Values
Most positron emission tomography (PET) emission scans are corrected for accidental coincidence (AC) events by real-time subtraction of delayed-window coincidences, leaving only the randoms-precorrected data available for image reconstruction. The real-time randoms precorrection compensates in mean for AC events but destroys the Poisson statistics. The exact log-likelihood for randoms-precorrected data is inconvenient, so practical approximations are needed for maximum likelihood or penalized-likelihood image reconstruction. Conventional approximations involve setting negative sinogram values to zero, which can induce positive systematic biases, particularly for scans with low counts per ray. We propose new likelihood approximations that allow negative sinogram values without requiring zero-thresholding. With negative sinogram values, the log-likelihood functions can be nonconcave, complicating maximization; nevertheless, we develop monotonic algorithms for the new models by modifying the separable paraboloidal surrogates and the maximum-likelihood expectation-maximization (ML-EM) methods. These algorithms ascend to local maximizers of the objective function. Analysis and simulation results show that the new shifted Poisson (SP) model is nearly free of systematic bias yet keeps low variance. Despite its simpler implementation, the new SP performs comparably to the saddle-point model which has shown the best performance (as to systematic bias and variance) in randoms-precorrected PET emission reconstruction.Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/85994/1/Fessler61.pd
Incremental Optimization Transfer Algorithms: Application to Transmission Tomography
No convergent ordered subsets (OS) type image
reconstruction algorithms for transmission tomography have been
proposed to date. In contrast, in emission tomography, there
are two known families of convergent OS algorithms: methods
that use relaxation parameters (Ahn and Fessler, 2003), and
methods based on the incremental expectation maximization (EM)
approach (Hsiao et al., 2002). This paper generalizes the incremental
EM approach by introducing a general framework that
we call “incremental optimization transfer.” Like incremental EM
methods, the proposed algorithms accelerate convergence speeds
and ensure global convergence (to a stationary point) under mild
regularity conditions without requiring inconvenient relaxation
parameters. The general optimization transfer framework enables
the use of a very broad family of non-EM surrogate functions.
In particular, this paper provides the first convergent OS-type
algorithm for transmission tomography. The general approach is
applicable to both monoenergetic and polyenergetic transmission
scans as well as to other image reconstruction problems. We
propose a particular incremental optimization transfer method
for (nonconcave) penalized-likelihood (PL) transmission image
reconstruction by using separable paraboloidal surrogates (SPS).
Results show that the new “transmission incremental optimization
transfer (TRIOT)” algorithm is faster than nonincremental
ordinary SPS and even OS-SPS yet is convergent.Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/85800/1/Fessler200.pd
Covariance of Kinetic Parameter Estimators Based on Time Activity Curve Reconstructions: Preliminary Study on 1D Dynamic Imaging
We provide approximate expressions for the covariance matrix of kinetic parameter estimators based on time activity curve (TAC) reconstructions when TACs are modeled as a linear combination of temporal basis functions such as B-splines. The approximations are useful tools for assessing and optimizing the basis functions for TACs and the temporal bins for data in terms of computation and efficiency. In this paper we analyze a 1D temporal problem for simplicity, and we consider a scenario where TACs are reconstructed by penalized-likelihood (PL) estimation incorporating temporal regularization, and kinetic parameters are obtained by maximum likelihood (ML) estimation. We derive approximate formulas for the covariance of the kinetic parameter estimators using 1) the mean and variance approximations for PL estimators in (Fessler, 1996) and 2) Cramer-Rao bounds. The approximations apply to list-mode data as well as bin-mode data.Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/85981/1/Fessler193.pd
- …